Classification Methods with Reject Option Based on Convex Risk Minimization
نویسندگان
چکیده
In this paper, we investigate the problem of binary classification with a reject option in which one can withhold the decision of classifying an observation at a cost lower than that of misclassification. Since the natural loss function is non-convex so that empirical risk minimization easily becomes infeasible, the paper proposes minimizing convex risks based on surrogate convex loss functions. A necessary and sufficient condition for infinite sample consistency (both risks share the same minimizer) is provided. Moreover, we show that the excess risk can be bounded through the excess surrogate risk under appropriate conditions. These bounds can be tightened by a generalized margin condition. The impact of the results is illustrated on several commonly used surrogate loss functions.
منابع مشابه
Lasso type classifiers with a reject option
This paper discusses structural risk minimization in the setting of classification with a reject option. Binary classification is about classifying observations that take values in an arbitrary feature space X into one of two classes, labelled −1 or +1. A discriminant function f : X → R yields a classifier sgn(f(x)) ∈ {−1,+1} that represents our guess of the label Y of a future observation X an...
متن کاملDouble Ramp Loss Based Reject Option Classifier
The performance of a reject option classifiers is quantified using 0 − d − 1 loss where d ∈ (0, .5) is the loss for rejection. In this paper, we propose double ramp loss function which gives a continuous upper bound for (0 − d − 1) loss. Our approach is based on minimizing regularized risk under the double ramp loss using difference of convex programming. We show the effectiveness of our approa...
متن کاملConsistency of Multiclass Empirical Risk Minimization Methods Based on Convex Loss
The consistency of classification algorithm plays a central role in statistical learning theory. A consistent algorithm guarantees us that taking more samples essentially suffices to roughly reconstruct the unknown distribution. We consider the consistency of ERM scheme over classes of combinations of very simple rules (base classifiers) in multiclass classification. Our approach is, under some...
متن کاملA Projected Alternating Least square Approach for Computation of Nonnegative Matrix Factorization
Nonnegative matrix factorization (NMF) is a common method in data mining that have been used in different applications as a dimension reduction, classification or clustering method. Methods in alternating least square (ALS) approach usually used to solve this non-convex minimization problem. At each step of ALS algorithms two convex least square problems should be solved, which causes high com...
متن کاملEfficiency of Classification Methods Based on Empirical Risk Minimization
EFFICIENCY OF CLASSIFICATION METHODS BASED ON EMPIRICAL RISK MINIMIZATION V. I. Norkin a and M. A. Keyzer b UDC 519:234:24:85 A binary classification problem is reduced to the minimization of convex regularized empirical risk functionals in a reproducing kernel Hilbert space. The solution is searched for in the form of a finite linear combination of kernel support functions (Vapnik’s support ve...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Journal of Machine Learning Research
دوره 11 شماره
صفحات -
تاریخ انتشار 2010